# f16 embedding optimization
Qwen3 8B GGUF
MIT
ZeroWw is a quantized text generation model that uses f16 format for output and embedding tensors, while other tensors use q5_k or q6_k format, resulting in a smaller size with performance comparable to pure f16.
Large Language Model English
Q
ZeroWw
236
1
Qwen3 4B GGUF
MIT
A quantized text generation model with output and embedding tensors in f16 format, while other tensors use q5_k or q6_k quantization, resulting in a smaller size with performance comparable to the pure f16 version.
Large Language Model English
Q
ZeroWw
495
2
Gemma 3 4b It Abliterated GGUF
MIT
An innovative quantization solution that achieves smaller model size while maintaining high performance through mixed-precision quantization.
Large Language Model English
G
ZeroWw
247
4
Featured Recommended AI Models